Algorithmic Differentiation of Numerical Methods: Second-Order Tangent and Adjoint Solvers for Systems of Parametrized Nonlinear Equations
نویسندگان
چکیده
Forward and reverse modes of algorithmic differentiation (AD) transform implementations of multivariate vector functions F : IR → IR as computer programs into tangent and adjoint code, respectively. The reapplication of the same ideas yields higher derivative code. In particular, second derivatives play an important role in nonlinear programming. Second-order methods based on Newton’s algorithm promise faster convergence in the neighbourhood of the minimum by taking into account second derivative information. The adjoint mode is of particular interest in large-scale gradient-based nonlinear optimization due to the independence of its computational cost on the number of free variables. Solvers for parametrized systems of n equations embedded into the evaluation of the objective function for a (without loss of generality) unconstrained nonlinear optimization problem require the Hessian of the objective with respect to the free variables implying the need for second derivatives of the nonlinear solver. The local computational overhead as well as the additional memory requirement for the computation of second-order tangents or second-order adjoints of the solution vector with respect to parameters by a fully algorithmic method (derived by AD) can quickly become prohibitive for large values of n. Both can be reduced significantly by the second-order symbolic approach to differentiation of the underlying numerical method to be discussed in this paper.
منابع مشابه
Algorithmic Differentiation of Numerical Methods: Tangent-Linear and Adjoint Direct Solvers for Systems of Linear Equations
We consider the Algorithmic Differentiation (also know as Automatic Differentiation; AD) of numerical simulation programs that contain calls to direct solvers for systems of n linear equations. AD of the linear solvers yields a local overhead of O(n) for the computation of directional derivatives or adjoints of the solution vector with respect to the system matrix and right-hand side. The local...
متن کاملAlgorithmic Differentiation of Numerical Methods: Tangent-Linear and Adjoint Solvers for Systems of Nonlinear Equations
We discuss software tool support for the Algorithmic Differentiation (also known as Automatic Differentiation; AD) of numerical simulation programs that contain calls to solvers for parameterized systems of n nonlinear equations. The local computational overhead as well as the additional memory requirement for the computation of directional derivatives or adjoints of the solution of the nonline...
متن کاملOn second derivative 3-stage Hermite--Birkhoff--Obrechkoff methods for stiff ODEs: A-stable up to order 10 with variable stepsize
Variable-step (VS) second derivative $k$-step $3$-stage Hermite--Birkhoff--Obrechkoff (HBO) methods of order $p=(k+3)$, denoted by HBO$(p)$ are constructed as a combination of linear $k$-step methods of order $(p-2)$ and a second derivative two-step diagonally implicit $3$-stage Hermite--Birkhoff method of order 5 (DIHB5) for solving stiff ordinary differential equations. The main reason for co...
متن کاملNew Solutions for Singular Lane-Emden Equations Arising in Astrophysics Based on Shifted Ultraspherical Operational Matrices of Derivatives
In this paper, the ultraspherical operational matrices of derivatives are constructed. Based on these operational matrices, two numerical algorithms are presented and analyzed for obtaining new approximate spectral solutions of a class of linear and nonlinear Lane-Emden type singular initial value problems. The basic idea behind the suggested algorithms is basically built on transforming the eq...
متن کاملForward and adjoint sensitivity analysis with continuous explicit Runge - Kutta methods ⋆
We study the numerical solution of tangent linear, first, and second order adjoint models with high-order explicit, continuous Runge-Kutta methods. The approaches currently implemented in popular packages such as SUNDIALS or DASPKADJOINT are based on linear multistep methods. For adaptive time integration of nonlinear models, interpolation of the forward model solution is required during the ad...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2014